2025-06-13 14:18:23,834 [ 172701 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-06-13 14:18:23,834 [ 172701 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:79, check_args_and_update_paths) 2025-06-13 14:18:23,834 [ 172701 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:90, check_args_and_update_paths) 2025-06-13 14:18:23,834 [ 172701 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:92, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_typopj --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests test_mask_sensitive_info/test.py::test_create_table -vvv " altinityinfra/integration-tests-runner:ad96270260ff '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [1 item] scheduling tests via LoadFileScheduling test_mask_sensitive_info/test.py::test_create_table [gw0] [100%] FAILED test_mask_sensitive_info/test.py::test_create_table =================================== FAILURES =================================== ______________________________ test_create_table _______________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 def test_create_table(): password = new_password() azure_conn_string = cluster.env_variables["AZURITE_CONNECTION_STRING"] account_key_pattern = re.compile("AccountKey=.*?(;|$)") masked_azure_conn_string = re.sub( account_key_pattern, "AccountKey=[HIDDEN]\\1", azure_conn_string ) masked_sas_conn_string = 'BlobEndpoint=https://my-endpoint/;SharedAccessSignature=[HIDDEN]' azure_storage_account_url = cluster.env_variables["AZURITE_STORAGE_ACCOUNT_URL"] azure_account_name = "devstoreaccount1" azure_account_key = "Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==" table_engines = [ f"MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '{password}')", f"PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '{password}')", f"MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '{password}')", f"S3('http://minio1:9001/root/data/test1.csv')", f"S3('http://minio1:9001/root/data/test2.csv', 'CSV')", f"S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip')", f"S3('http://minio1:9001/root/data/test4.csv', 'minio', '{password}', 'CSV')", f"S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', '{password}', 'CSV', 'gzip')", f"MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', table = 'mysql_table', user = 'mysql_user', password = '{password}')", f"MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = '{password}', table = 'mysql_table', user = 'mysql_user')", f"MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table')", f"PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', table = 'postgres_table', user = 'postgres_user', password = '{password}')", f"MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = '{password}')", f"S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = '{password}', format = 'CSV')", f"S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip')", f"S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '{password}')", ( f"DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), f"S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered'", f"S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered'", f"S3Queue('http://minio1:9001/root/data/', 'minio', '{password}', 'CSV') settings mode = 'ordered'", f"S3Queue('http://minio1:9001/root/data/', 'minio', '{password}', 'CSV', 'gzip') settings mode = 'ordered'", ( f"Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), ( f"IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), ( f"Iceberg(storage_type='s3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), f"AzureBlobStorage('{azure_conn_string}', 'cont', 'test_simple.csv', 'CSV')", f"AzureBlobStorage('{azure_conn_string}', 'cont', 'test_simple_1.csv', 'CSV', 'none')", f"AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_3.csv', '{azure_account_name}', '{azure_account_key}')", f"AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_4.csv', '{azure_account_name}', '{azure_account_key}', 'CSV')", f"AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_5.csv', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none')", f"AzureQueue('{azure_conn_string}', 'cont', '*', 'CSV') SETTINGS mode = 'unordered'", f"AzureQueue('{azure_conn_string}', 'cont', '*', 'CSV', 'none') SETTINGS mode = 'unordered'", f"AzureQueue('{azure_storage_account_url}', 'cont', '*', '{azure_account_name}', '{azure_account_key}', 'CSV') SETTINGS mode = 'unordered'", f"AzureQueue('{azure_storage_account_url}', 'cont', '*', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none') SETTINGS mode = 'unordered'", ( f"AzureBlobStorage('BlobEndpoint=https://my-endpoint/;SharedAccessSignature=sp=r&st=2025-09-29T14:58:11Z&se=2025-09-29T00:00:00Z&spr=https&sv=2022-11-02&sr=c&sig=SECRET%SECRET%SECRET%SECRET', 'exampledatasets', 'example.csv')", "STD_EXCEPTION", ), f"AzureBlobStorage(named_collection_2, connection_string = '{azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", f"AzureBlobStorage(named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '{azure_account_key}')", ( f"IcebergAzure('{azure_conn_string}', 'cont', 'test_simple.csv')", "FILE_DOESNT_EXIST", ), ( f"IcebergAzure('{azure_storage_account_url}', 'cont', 'test_simple.csv', '{azure_account_name}', '{azure_account_key}')", "FILE_DOESNT_EXIST", ), ( f"IcebergAzure(named_collection_2, connection_string = '{azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", "FILE_DOESNT_EXIST", ), ( f"IcebergAzure(named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '{azure_account_key}')", "FILE_DOESNT_EXIST", ), ( f"Iceberg(storage_type='azure', '{azure_conn_string}', 'cont', 'test_simple.csv')", "FILE_DOESNT_EXIST", ), ( f"Iceberg(storage_type='azure', '{azure_storage_account_url}', 'cont', 'test_simple.csv', '{azure_account_name}', '{azure_account_key}')", "FILE_DOESNT_EXIST", ), ( f"Iceberg(storage_type='azure', named_collection_2, connection_string = '{azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", "FILE_DOESNT_EXIST", ), ( f"Iceberg(storage_type='azure', named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '{azure_account_key}')", "FILE_DOESNT_EXIST", ), ] def make_test_case(i): table_name = f"table{i}" table_engine = table_engines[i] error = None if isinstance(table_engine, tuple): table_engine, error = table_engine query = f"CREATE TABLE {table_name} (x int) ENGINE = {table_engine}" return table_name, query, error # Generate test cases as a list of tuples (table_name, query, error). test_cases = [make_test_case(i) for i in range(len(table_engines))] for table_name, query, error in test_cases: if error: assert error in node.query_and_get_error(query) else: node.query(query) for toggle, secret in enumerate(["[HIDDEN]", password]): assert ( node.query(f"SHOW CREATE TABLE table0 {show_secrets}={toggle}") == "CREATE TABLE default.table0\\n(\\n `x` Int32\\n)\\n" "ENGINE = MySQL(\\'mysql80:3306\\', \\'mysql_db\\', " f"\\'mysql_table\\', \\'mysql_user\\', \\'{secret}\\')\n" ) assert node.query( f"SELECT create_table_query, engine_full FROM system.tables WHERE name = 'table0' {show_secrets}={toggle}" ) == TSV( [ [ "CREATE TABLE default.table0 (`x` Int32) ENGINE = MySQL(\\'mysql80:3306\\', \\'mysql_db\\', " f"\\'mysql_table\\', \\'mysql_user\\', \\'{secret}\\')", f"MySQL(\\'mysql80:3306\\', \\'mysql_db\\', \\'mysql_table\\', \\'mysql_user\\', \\'{secret}\\')", ], ] ) > check_logs( must_contain=[ "CREATE TABLE table0 (`x` int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '[HIDDEN]')", "CREATE TABLE table1 (`x` int) ENGINE = PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '[HIDDEN]')", "CREATE TABLE table2 (`x` int) ENGINE = MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '[HIDDEN]')", "CREATE TABLE table3 (x int) ENGINE = S3('http://minio1:9001/root/data/test1.csv')", "CREATE TABLE table4 (x int) ENGINE = S3('http://minio1:9001/root/data/test2.csv', 'CSV')", "CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip')", "CREATE TABLE table6 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test4.csv', 'minio', '[HIDDEN]', 'CSV')", "CREATE TABLE table7 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', '[HIDDEN]', 'CSV', 'gzip')", "CREATE TABLE table8 (`x` int) ENGINE = MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', `table` = 'mysql_table', user = 'mysql_user', password = '[HIDDEN]')", "CREATE TABLE table9 (`x` int) ENGINE = MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = '[HIDDEN]', `table` = 'mysql_table', user = 'mysql_user')", "CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table')", "CREATE TABLE table11 (`x` int) ENGINE = PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', `table` = 'postgres_table', user = 'postgres_user', password = '[HIDDEN]')", "CREATE TABLE table12 (`x` int) ENGINE = MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = '[HIDDEN]'", "CREATE TABLE table13 (`x` int) ENGINE = S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = '[HIDDEN]', format = 'CSV')", "CREATE TABLE table14 (x int) ENGINE = S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip')", "CREATE TABLE table15 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '[HIDDEN]')", "CREATE TABLE table16 (`x` int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')", "CREATE TABLE table17 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered'", "CREATE TABLE table18 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered'", # due to sensitive data substituion the query will be normalized, so not "settings" but "SETTINGS" "CREATE TABLE table19 (`x` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '[HIDDEN]', 'CSV') SETTINGS mode = 'ordered'", "CREATE TABLE table20 (`x` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '[HIDDEN]', 'CSV', 'gzip') SETTINGS mode = 'ordered'", "CREATE TABLE table21 (`x` int) ENGINE = Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')", "CREATE TABLE table22 (`x` int) ENGINE = IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')", "CREATE TABLE table23 (`x` int) ENGINE = Iceberg(storage_type = 's3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')", f"CREATE TABLE table24 (`x` int) ENGINE = AzureBlobStorage('{masked_azure_conn_string}', 'cont', 'test_simple.csv', 'CSV')", f"CREATE TABLE table25 (`x` int) ENGINE = AzureBlobStorage('{masked_azure_conn_string}', 'cont', 'test_simple_1.csv', 'CSV', 'none')", f"CREATE TABLE table26 (`x` int) ENGINE = AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_3.csv', '{azure_account_name}', '[HIDDEN]')", f"CREATE TABLE table27 (`x` int) ENGINE = AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_4.csv', '{azure_account_name}', '[HIDDEN]', 'CSV')", f"CREATE TABLE table28 (`x` int) ENGINE = AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_5.csv', '{azure_account_name}', '[HIDDEN]', 'CSV', 'none')", f"CREATE TABLE table29 (`x` int) ENGINE = AzureBlobStorage(named_collection_2, connection_string = '{masked_azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", f"CREATE TABLE table30 (`x` int) ENGINE = AzureBlobStorage(named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '[HIDDEN]')", f"CREATE TABLE table31 (`x` int) ENGINE = IcebergAzure('{masked_azure_conn_string}', 'cont', 'test_simple.csv')", f"CREATE TABLE table32 (`x` int) ENGINE = IcebergAzure('{azure_storage_account_url}', 'cont', 'test_simple.csv', '{azure_account_name}', '[HIDDEN]')", f"CREATE TABLE table33 (`x` int) ENGINE = IcebergAzure(named_collection_2, connection_string = '{masked_azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", f"CREATE TABLE table34 (`x` int) ENGINE = IcebergAzure(named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '[HIDDEN]')", f"CREATE TABLE table35 (`x` int) ENGINE = Iceberg(storage_type = 'azure', '{masked_azure_conn_string}', 'cont', 'test_simple.csv')", f"CREATE TABLE table36 (`x` int) ENGINE = Iceberg(storage_type = 'azure', '{azure_storage_account_url}', 'cont', 'test_simple.csv', '{azure_account_name}', '[HIDDEN]')", f"CREATE TABLE table37 (`x` int) ENGINE = Iceberg(storage_type = 'azure', named_collection_2, connection_string = '{masked_azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", f"CREATE TABLE table38 (`x` int) ENGINE = Iceberg(storage_type = 'azure', named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '[HIDDEN]')", f"CREATE TABLE table28 (`x` int) ENGINE = AzureQueue('{masked_azure_conn_string}', 'cont', '*', 'CSV') SETTINGS mode = 'unordered'", f"CREATE TABLE table29 (`x` int) ENGINE = AzureQueue('{masked_azure_conn_string}', 'cont', '*', 'CSV', 'none') SETTINGS mode = 'unordered'", f"CREATE TABLE table30 (`x` int) ENGINE = AzureQueue('{azure_storage_account_url}', 'cont', '*', '{azure_account_name}', '[HIDDEN]', 'CSV') SETTINGS mode = 'unordered'", f"CREATE TABLE table31 (`x` int) ENGINE = AzureQueue('{azure_storage_account_url}', 'cont', '*', '{azure_account_name}', '[HIDDEN]', 'CSV', 'none') SETTINGS mode = 'unordered'", f"CREATE TABLE table32 (`x` int) ENGINE = AzureBlobStorage('{masked_sas_conn_string}', 'exampledatasets', 'example.csv')", ], must_not_contain=[password], ) test_mask_sensitive_info/test.py:369: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ must_contain = ["CREATE TABLE table0 (`x` int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '[HIDDEN]')", ...', 'CSV')", "CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip')", ...] must_not_contain = ['VAYUI90LWUBA7PDG'] def check_logs(must_contain=[], must_not_contain=[]): node.query("SYSTEM FLUSH LOGS") for str in must_contain: escaped_str = ( str.replace("`", "\\`") .replace("[", "\\[") .replace("]", "\\]") .replace("*", "\\*") ) > assert node.contains_in_log(escaped_str, exclusion_substring=base_search_query) E assert False E + where False = >("CREATE TABLE table29 (\\`x\\` int) ENGINE = AzureBlobStorage(named_collection_2, connection_string = 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\\[HIDDEN\\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", exclusion_substring='SELECT COUNT() FROM system.query_log WHERE query LIKE ') E + where > = .contains_in_log test_mask_sensitive_info/test.py:43: AssertionError ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2025-06-13 14:18:30 [ 671 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2025-06-13 14:18:30 [ 671 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2025-06-13 14:18:30 [ 671 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2025-06-13 14:18:30 [ 671 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2025-06-13 14:18:30 [ 671 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Stdout:Deleted Networks: (cluster.py:145, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Stdout:roottestmysqlprotocol-gw1_default (cluster.py:145, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Stdout: (cluster.py:145, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2025-06-13 14:18:31 [ 671 ] INFO : Running tests in /ClickHouse/tests/integration/test_mask_sensitive_info/test.py (cluster.py:2672, start) 2025-06-13 14:18:31 [ 671 ] DEBUG : Cluster start called. is_up=False (cluster.py:2679, start) 2025-06-13 14:18:31 [ 671 ] DEBUG : Docker networks for project roottestmasksensitiveinfo-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:825, print_all_docker_pieces) 2025-06-13 14:18:31 [ 671 ] DEBUG : Docker containers for project roottestmasksensitiveinfo-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:833, print_all_docker_pieces) 2025-06-13 14:18:31 [ 671 ] DEBUG : Docker volumes for project roottestmasksensitiveinfo-gw0 are DRIVER VOLUME NAME (cluster.py:841, print_all_docker_pieces) 2025-06-13 14:18:31 [ 671 ] DEBUG : Cleanup called (cluster.py:846, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Docker networks for project roottestmasksensitiveinfo-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:825, print_all_docker_pieces) 2025-06-13 14:18:31 [ 671 ] DEBUG : Docker containers for project roottestmasksensitiveinfo-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:833, print_all_docker_pieces) 2025-06-13 14:18:31 [ 671 ] DEBUG : Docker volumes for project roottestmasksensitiveinfo-gw0 are DRIVER VOLUME NAME (cluster.py:841, print_all_docker_pieces) 2025-06-13 14:18:31 [ 671 ] DEBUG : Command:[docker container list --all --filter name='^/roottestmasksensitiveinfo-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Unstopped containers: {} (cluster.py:860, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : No running containers for project: roottestmasksensitiveinfo-gw0 (cluster.py:874, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Trying to prune unused networks... (cluster.py:880, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Trying to prune unused images... (cluster.py:896, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Images pruned (cluster.py:899, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Trying to prune unused volumes... (cluster.py:905, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2025-06-13 14:18:31 [ 671 ] DEBUG : Volumes pruned: 1 (cluster.py:910, cleanup) 2025-06-13 14:18:31 [ 671 ] DEBUG : Setup directory for instance: node (cluster.py:2692, start) 2025-06-13 14:18:31 [ 671 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4536, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Create directory for common tests configuration (cluster.py:4541, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Copy common configuration from helpers (cluster.py:4561, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Generate and write macros file (cluster.py:4613, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_mask_sensitive_info/configs/overrides.xml'] to /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/configs/config.d (cluster.py:4649, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/database (cluster.py:4666, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/logs (cluster.py:4677, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4758, create_dir) 2025-06-13 14:18:31 [ 671 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:5ccda723c1fc', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/coordination', 'AZURITE_PORT': '30000', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30000/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env (cluster.py:96, _create_env_file) 2025-06-13 14:18:31 [ 671 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-06-13 14:18:31 [ 671 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-06-13 14:18:31 [ 671 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-06-13 14:18:31 [ 671 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-06-13 14:18:31 [ 671 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-06-13 14:18:31 [ 671 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml pull] (cluster.py:121, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by zoo2 (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: node Skipped - Image is already being pulled by zoo2 (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by zoo2 (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: azurite1 Pulling (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: zoo2 Pulling (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: zoo2 Pulled (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Stderr: azurite1 Pulled (cluster.py:147, run_and_check) 2025-06-13 14:19:32 [ 671 ] DEBUG : Setup ZooKeeper (cluster.py:2733, start) 2025-06-13 14:19:32 [ 671 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/log', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/config', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/log', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/config', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/log', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/config', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/coordination'] (cluster.py:2734, start) 2025-06-13 14:19:32 [ 671 ] DEBUG : Command:[docker compose --project-name roottestmasksensitiveinfo-gw0 --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:121, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:32Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Created (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Created (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Created (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Created (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Starting (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Starting (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Starting (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Started (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Started (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Started (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:33Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:33Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2025-06-13 14:19:33 [ 671 ] DEBUG : Wait ZooKeeper to start (cluster.py:2398, wait_zookeeper_to_start) 2025-06-13 14:19:33 [ 671 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:1999, get_instance_ip) 2025-06-13 14:19:33 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:33 [ 671 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.1.4, port:2181, use_ssl:False (cluster.py:3234, get_kazoo_client) 2025-06-13 14:19:33 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:33 [ 671 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-06-13 14:19:33 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:33 [ 671 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-06-13 14:19:33 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:33 [ 671 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-06-13 14:19:34 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:34 [ 671 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-06-13 14:19:34 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:34 [ 671 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-06-13 14:19:35 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:35 [ 671 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-06-13 14:19:37 [ 671 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-06-13 14:19:37 [ 671 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:1999, get_instance_ip) 2025-06-13 14:19:37 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:37 [ 671 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.1.2, port:2181, use_ssl:False (cluster.py:3234, get_kazoo_client) 2025-06-13 14:19:37 [ 671 ] INFO : Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-06-13 14:19:37 [ 671 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-06-13 14:19:37 [ 671 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:1999, get_instance_ip) 2025-06-13 14:19:37 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:37 [ 671 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.1.3, port:2181, use_ssl:False (cluster.py:3234, get_kazoo_client) 2025-06-13 14:19:37 [ 671 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-06-13 14:19:37 [ 671 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-06-13 14:19:37 [ 671 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-06-13 14:19:37 [ 671 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-06-13 14:19:37 [ 671 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-06-13 14:19:37 [ 671 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2414, wait_zookeeper_nodes_to_start) 2025-06-13 14:19:37 [ 671 ] INFO : Trying to create Azurite instance by command docker compose --project-name roottestmasksensitiveinfo-gw0 --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d (cluster.py:3004, start) 2025-06-13 14:19:37 [ 671 ] DEBUG : Command:[docker compose --project-name roottestmasksensitiveinfo-gw0 --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] (cluster.py:121, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:37Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr: Volume "roottestmasksensitiveinfo-gw0_data1-1" Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr: Volume "roottestmasksensitiveinfo-gw0_data1-1" Created (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:37Z" level=warning msg="Found orphan containers ([roottestmasksensitiveinfo-gw0-zoo2-1 roottestmasksensitiveinfo-gw0-zoo3-1 roottestmasksensitiveinfo-gw0-zoo1-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Created (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Starting (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Started (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:38Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] DEBUG : Stderr:time="2025-06-13T14:19:38Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2025-06-13 14:19:38 [ 671 ] INFO : Trying to connect to Azurite (cluster.py:3018, start) 2025-06-13 14:19:39 [ 671 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '70f584ce-4861-11f0-840e-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : Starting new HTTP connection (1): 127.0.0.1:30000 (connectionpool.py:245, _new_conn) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://127.0.0.1:30000 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] INFO : Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '70f584ce-4861-11f0-840e-0242ac110002' 'x-ms-request-id': '1f044487-f5d4-49d7-ac09-79ce18a7e393' 'x-ms-version': 'REDACTED' 'date': 'Fri, 13 Jun 2025 14:19:39 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' (_universal.py:550, on_response) 2025-06-13 14:19:39 [ 671 ] DEBUG : {'client_request_id': '70f584ce-4861-11f0-840e-0242ac110002', 'request_id': '1f044487-f5d4-49d7-ac09-79ce18a7e393', 'version': '2025-05-05', 'date': datetime.datetime(2025, 6, 13, 14, 19, 39, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} (cluster.py:2544, wait_azurite_to_start) 2025-06-13 14:19:39 [ 671 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '70fdc4b8-4861-11f0-840e-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://127.0.0.1:30000 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] INFO : Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '70fdc4b8-4861-11f0-840e-0242ac110002' 'x-ms-request-id': 'd613d6d6-b79d-456a-88f3-03022359038f' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Fri, 13 Jun 2025 14:19:39 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' (_universal.py:550, on_response) 2025-06-13 14:19:39 [ 671 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '70ff98b0-4861-11f0-840e-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://127.0.0.1:30000 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] INFO : Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': 'd3d7b01c-e80a-408a-8511-8c9c0784c5e7' 'content-type': 'application/xml' 'Date': 'Fri, 13 Jun 2025 14:19:39 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' (_universal.py:550, on_response) 2025-06-13 14:19:39 [ 671 ] DEBUG : azurite container 'azurite-container' doesn't exist, creating it (cluster.py:2566, wait_azurite_to_start) 2025-06-13 14:19:39 [ 671 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '7100f3e0-4861-11f0-840e-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://127.0.0.1:30000 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] INFO : Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'etag': '"0x1F924F0BB522650"' 'last-modified': 'Fri, 13 Jun 2025 14:19:39 GMT' 'x-ms-client-request-id': '7100f3e0-4861-11f0-840e-0242ac110002' 'x-ms-request-id': '3eb1199a-8adf-401c-8974-f153d58f79a9' 'x-ms-version': 'REDACTED' 'Date': 'Fri, 13 Jun 2025 14:19:39 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' (_universal.py:550, on_response) 2025-06-13 14:19:39 [ 671 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml up -d --no-recreate') (cluster.py:3061, start) 2025-06-13 14:19:39 [ 671 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Running (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Running (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Running (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Running (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Creating (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Created (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Starting (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Started (cluster.py:147, run_and_check) 2025-06-13 14:19:39 [ 671 ] DEBUG : ClickHouse instance created (cluster.py:3069, start) 2025-06-13 14:19:39 [ 671 ] DEBUG : get_instance_ip instance_name=node (cluster.py:1999, get_instance_ip) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2009, get_instance_global_ipv6) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : Waiting for ClickHouse start in node, ip: 172.16.1.6... (cluster.py:3077, start) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:39 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:40 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:41 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:41 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:41 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:41 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:41 [ 671 ] DEBUG : http://localhost:None "GET /v1.46/containers/0029865756c8940cb63e1da4767f92af0f6d69f09e2730d5e55608bad0e79f11/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-06-13 14:19:41 [ 671 ] DEBUG : ClickHouse node started (cluster.py:3081, start) ------------------------------ Captured log call ------------------------------- 2025-06-13 14:19:41 [ 671 ] DEBUG : Executing query CREATE TABLE table0 (x int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:41 [ 671 ] DEBUG : Executing query CREATE TABLE table1 (x int) ENGINE = PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:41 [ 671 ] DEBUG : Executing query CREATE TABLE table2 (x int) ENGINE = MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:42 [ 671 ] DEBUG : Executing query CREATE TABLE table3 (x int) ENGINE = S3('http://minio1:9001/root/data/test1.csv') on node (cluster.py:3570, query) 2025-06-13 14:19:42 [ 671 ] DEBUG : Executing query CREATE TABLE table4 (x int) ENGINE = S3('http://minio1:9001/root/data/test2.csv', 'CSV') on node (cluster.py:3570, query) 2025-06-13 14:19:42 [ 671 ] DEBUG : Executing query CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip') on node (cluster.py:3570, query) 2025-06-13 14:19:43 [ 671 ] DEBUG : Executing query CREATE TABLE table6 (x int) ENGINE = S3('http://minio1:9001/root/data/test4.csv', 'minio', 'VAYUI90LWUBA7PDG', 'CSV') on node (cluster.py:3570, query) 2025-06-13 14:19:43 [ 671 ] DEBUG : Executing query CREATE TABLE table7 (x int) ENGINE = S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', 'VAYUI90LWUBA7PDG', 'CSV', 'gzip') on node (cluster.py:3570, query) 2025-06-13 14:19:43 [ 671 ] DEBUG : Executing query CREATE TABLE table8 (x int) ENGINE = MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', table = 'mysql_table', user = 'mysql_user', password = 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:43 [ 671 ] DEBUG : Executing query CREATE TABLE table9 (x int) ENGINE = MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = 'VAYUI90LWUBA7PDG', table = 'mysql_table', user = 'mysql_user') on node (cluster.py:3570, query) 2025-06-13 14:19:44 [ 671 ] DEBUG : Executing query CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table') on node (cluster.py:3570, query) 2025-06-13 14:19:44 [ 671 ] DEBUG : Executing query CREATE TABLE table11 (x int) ENGINE = PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', table = 'postgres_table', user = 'postgres_user', password = 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:44 [ 671 ] DEBUG : Executing query CREATE TABLE table12 (x int) ENGINE = MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:44 [ 671 ] DEBUG : Executing query CREATE TABLE table13 (x int) ENGINE = S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = 'VAYUI90LWUBA7PDG', format = 'CSV') on node (cluster.py:3570, query) 2025-06-13 14:19:45 [ 671 ] DEBUG : Executing query CREATE TABLE table14 (x int) ENGINE = S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip') on node (cluster.py:3570, query) 2025-06-13 14:19:45 [ 671 ] DEBUG : Executing query CREATE TABLE table15 (x int) ENGINE = S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', 'VAYUI90LWUBA7PDG') on node (cluster.py:3570, query) 2025-06-13 14:19:45 [ 671 ] DEBUG : Executing query CREATE TABLE table16 (x int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'VAYUI90LWUBA7PDG') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:20:00 [ 671 ] DEBUG : Executing query CREATE TABLE table17 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered' on node (cluster.py:3570, query) 2025-06-13 14:20:00 [ 671 ] DEBUG : Executing query CREATE TABLE table18 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered' on node (cluster.py:3570, query) 2025-06-13 14:20:01 [ 671 ] DEBUG : Executing query CREATE TABLE table19 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', 'VAYUI90LWUBA7PDG', 'CSV') settings mode = 'ordered' on node (cluster.py:3570, query) 2025-06-13 14:20:01 [ 671 ] DEBUG : Executing query CREATE TABLE table20 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', 'VAYUI90LWUBA7PDG', 'CSV', 'gzip') settings mode = 'ordered' on node (cluster.py:3570, query) 2025-06-13 14:20:01 [ 671 ] DEBUG : Executing query CREATE TABLE table21 (x int) ENGINE = Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'VAYUI90LWUBA7PDG') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:20:15 [ 671 ] DEBUG : Executing query CREATE TABLE table22 (x int) ENGINE = IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'VAYUI90LWUBA7PDG') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:20:29 [ 671 ] DEBUG : Executing query CREATE TABLE table23 (x int) ENGINE = Iceberg(storage_type='s3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', 'VAYUI90LWUBA7PDG') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:20:44 [ 671 ] DEBUG : Executing query CREATE TABLE table24 (x int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple.csv', 'CSV') on node (cluster.py:3570, query) 2025-06-13 14:20:45 [ 671 ] DEBUG : Executing query CREATE TABLE table25 (x int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple_1.csv', 'CSV', 'none') on node (cluster.py:3570, query) 2025-06-13 14:20:45 [ 671 ] DEBUG : Executing query CREATE TABLE table26 (x int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_3.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node (cluster.py:3570, query) 2025-06-13 14:20:45 [ 671 ] DEBUG : Executing query CREATE TABLE table27 (x int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_4.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV') on node (cluster.py:3570, query) 2025-06-13 14:20:46 [ 671 ] DEBUG : Executing query CREATE TABLE table28 (x int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_5.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'none') on node (cluster.py:3570, query) 2025-06-13 14:20:46 [ 671 ] DEBUG : Executing query CREATE TABLE table29 (x int) ENGINE = AzureQueue('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', '*', 'CSV') SETTINGS mode = 'unordered' on node (cluster.py:3570, query) 2025-06-13 14:20:46 [ 671 ] DEBUG : Executing query CREATE TABLE table30 (x int) ENGINE = AzureQueue('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', '*', 'CSV', 'none') SETTINGS mode = 'unordered' on node (cluster.py:3570, query) 2025-06-13 14:20:46 [ 671 ] DEBUG : Executing query CREATE TABLE table31 (x int) ENGINE = AzureQueue('http://azurite1:30000/devstoreaccount1', 'cont', '*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV') SETTINGS mode = 'unordered' on node (cluster.py:3570, query) 2025-06-13 14:20:47 [ 671 ] DEBUG : Executing query CREATE TABLE table32 (x int) ENGINE = AzureQueue('http://azurite1:30000/devstoreaccount1', 'cont', '*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'none') SETTINGS mode = 'unordered' on node (cluster.py:3570, query) 2025-06-13 14:20:47 [ 671 ] DEBUG : Executing query CREATE TABLE table33 (x int) ENGINE = AzureBlobStorage('BlobEndpoint=https://my-endpoint/;SharedAccessSignature=sp=r&st=2025-09-29T14:58:11Z&se=2025-09-29T00:00:00Z&spr=https&sv=2022-11-02&sr=c&sig=SECRET%SECRET%SECRET%SECRET', 'exampledatasets', 'example.csv') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:14 [ 671 ] DEBUG : Executing query CREATE TABLE table34 (x int) ENGINE = AzureBlobStorage(named_collection_2, connection_string = 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV') on node (cluster.py:3570, query) 2025-06-13 14:21:14 [ 671 ] DEBUG : Executing query CREATE TABLE table35 (x int) ENGINE = AzureBlobStorage(named_collection_2, storage_account_url = 'http://azurite1:30000/devstoreaccount1', container = 'cont', blob_path = 'test_simple_8.csv', account_name = 'devstoreaccount1', account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node (cluster.py:3570, query) 2025-06-13 14:21:14 [ 671 ] DEBUG : Executing query CREATE TABLE table36 (x int) ENGINE = IcebergAzure('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple.csv') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:15 [ 671 ] DEBUG : Executing query CREATE TABLE table37 (x int) ENGINE = IcebergAzure('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:15 [ 671 ] DEBUG : Executing query CREATE TABLE table38 (x int) ENGINE = IcebergAzure(named_collection_2, connection_string = 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:15 [ 671 ] DEBUG : Executing query CREATE TABLE table39 (x int) ENGINE = IcebergAzure(named_collection_2, storage_account_url = 'http://azurite1:30000/devstoreaccount1', container = 'cont', blob_path = 'test_simple_8.csv', account_name = 'devstoreaccount1', account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:16 [ 671 ] DEBUG : Executing query CREATE TABLE table40 (x int) ENGINE = Iceberg(storage_type='azure', 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple.csv') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:16 [ 671 ] DEBUG : Executing query CREATE TABLE table41 (x int) ENGINE = Iceberg(storage_type='azure', 'http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:16 [ 671 ] DEBUG : Executing query CREATE TABLE table42 (x int) ENGINE = Iceberg(storage_type='azure', named_collection_2, connection_string = 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:17 [ 671 ] DEBUG : Executing query CREATE TABLE table43 (x int) ENGINE = Iceberg(storage_type='azure', named_collection_2, storage_account_url = 'http://azurite1:30000/devstoreaccount1', container = 'cont', blob_path = 'test_simple_8.csv', account_name = 'devstoreaccount1', account_key = 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node (cluster.py:3653, query_and_get_error) 2025-06-13 14:21:17 [ 671 ] DEBUG : Executing query SHOW CREATE TABLE table0 SETTINGS format_display_secrets_in_show_and_select=0 on node (cluster.py:3570, query) 2025-06-13 14:21:17 [ 671 ] DEBUG : Executing query SELECT create_table_query, engine_full FROM system.tables WHERE name = 'table0' SETTINGS format_display_secrets_in_show_and_select=0 on node (cluster.py:3570, query) 2025-06-13 14:21:18 [ 671 ] DEBUG : Executing query SHOW CREATE TABLE table0 SETTINGS format_display_secrets_in_show_and_select=1 on node (cluster.py:3570, query) 2025-06-13 14:21:18 [ 671 ] DEBUG : Executing query SELECT create_table_query, engine_full FROM system.tables WHERE name = 'table0' SETTINGS format_display_secrets_in_show_and_select=1 on node (cluster.py:3570, query) 2025-06-13 14:21:18 [ 671 ] DEBUG : Executing query SYSTEM FLUSH LOGS on node (cluster.py:3570, query) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table0 (\\`x\\` int) ENGINE = MySQL(\'mysql80:3306\', \'mysql_db\', \'mysql_table\', \'mysql_user\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table0 (\`x\` int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:41.600256 [ 9 ] {c608d4ef-db02-4ab0-b8c6-eaaadc0ae44f} executeQuery: (from 172.16.1.1:64290) (query 1, line 1) CREATE TABLE table0 (`x` int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:41.600382 [ 9 ] {c608d4ef-db02-4ab0-b8c6-eaaadc0ae44f} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table0 (`x` int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table1 (\\`x\\` int) ENGINE = PostgreSQL(\'postgres1:5432\', \'postgres_db\', \'postgres_table\', \'postgres_user\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table1 (\`x\` int) ENGINE = PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:41.850601 [ 9 ] {dd7eec14-5021-427f-bb45-105a9b515cda} executeQuery: (from 172.16.1.1:64302) (query 1, line 1) CREATE TABLE table1 (`x` int) ENGINE = PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:41.850767 [ 9 ] {dd7eec14-5021-427f-bb45-105a9b515cda} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table1 (`x` int) ENGINE = PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table2 (\\`x\\` int) ENGINE = MongoDB(\'mongo1:27017\', \'mongo_db\', \'mongo_col\', \'mongo_user\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table2 (\`x\` int) ENGINE = MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.061020 [ 9 ] {15946a4b-0caf-4cc6-a8b2-bb7fc0e3bf7a} executeQuery: (from 172.16.1.1:64308) (query 1, line 1) CREATE TABLE table2 (`x` int) ENGINE = MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.061192 [ 9 ] {15946a4b-0caf-4cc6-a8b2-bb7fc0e3bf7a} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table2 (`x` int) ENGINE = MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table3 (x int) ENGINE = S3(\'http://minio1:9001/root/data/test1.csv\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table3 (x int) ENGINE = S3('http://minio1:9001/root/data/test1.csv')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.305542 [ 9 ] {1f77478c-d93c-4927-a8ee-b36078e29415} executeQuery: (from 172.16.1.1:64316) (query 1, line 1) CREATE TABLE table3 (x int) ENGINE = S3('http://minio1:9001/root/data/test1.csv') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.305668 [ 9 ] {1f77478c-d93c-4927-a8ee-b36078e29415} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table3 (x int) ENGINE = S3('http://minio1:9001/root/data/test1.csv') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table4 (x int) ENGINE = S3(\'http://minio1:9001/root/data/test2.csv\', \'CSV\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table4 (x int) ENGINE = S3('http://minio1:9001/root/data/test2.csv', 'CSV')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.652674 [ 9 ] {f188f6b1-030d-4f28-88d1-7d5842e700c4} executeQuery: (from 172.16.1.1:64330) (query 1, line 1) CREATE TABLE table4 (x int) ENGINE = S3('http://minio1:9001/root/data/test2.csv', 'CSV') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.652835 [ 9 ] {f188f6b1-030d-4f28-88d1-7d5842e700c4} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table4 (x int) ENGINE = S3('http://minio1:9001/root/data/test2.csv', 'CSV') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table5 (x int) ENGINE = S3(\'http://minio1:9001/root/data/test3.csv.gz\', \'CSV\', \'gzip\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.909009 [ 9 ] {90718990-f258-4c00-93ca-865a679e9ca9} executeQuery: (from 172.16.1.1:64338) (query 1, line 1) CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:42.909130 [ 9 ] {90718990-f258-4c00-93ca-865a679e9ca9} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table6 (\\`x\\` int) ENGINE = S3(\'http://minio1:9001/root/data/test4.csv\', \'minio\', \'\\[HIDDEN\\]\', \'CSV\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table6 (\`x\` int) ENGINE = S3('http://minio1:9001/root/data/test4.csv', 'minio', '\[HIDDEN\]', 'CSV')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.178726 [ 9 ] {04d446a9-5415-4373-a892-643bbdd50ace} executeQuery: (from 172.16.1.1:64342) (query 1, line 1) CREATE TABLE table6 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test4.csv', 'minio', '[HIDDEN]', 'CSV') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.178824 [ 9 ] {04d446a9-5415-4373-a892-643bbdd50ace} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table6 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test4.csv', 'minio', '[HIDDEN]', 'CSV') (cluster.py:145, run_and_check) 2025-06-13 14:21:19 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table7 (\\`x\\` int) ENGINE = S3(\'http://minio1:9001/root/data/test5.csv.gz\', \'minio\', \'\\[HIDDEN\\]\', \'CSV\', \'gzip\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:19 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table7 (\`x\` int) ENGINE = S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', '\[HIDDEN\]', 'CSV', 'gzip')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.453157 [ 9 ] {1ed7fd17-0248-4cb4-a377-2fd496f228ec} executeQuery: (from 172.16.1.1:64356) (query 1, line 1) CREATE TABLE table7 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', '[HIDDEN]', 'CSV', 'gzip') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.453340 [ 9 ] {1ed7fd17-0248-4cb4-a377-2fd496f228ec} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table7 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', '[HIDDEN]', 'CSV', 'gzip') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table8 (\\`x\\` int) ENGINE = MySQL(named_collection_1, host = \'mysql80\', port = 3306, database = \'mysql_db\', \\`table\\` = \'mysql_table\', user = \'mysql_user\', password = \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table8 (\`x\` int) ENGINE = MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', \`table\` = 'mysql_table', user = 'mysql_user', password = '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.711230 [ 9 ] {44018ed3-7ebe-473a-87d9-50d699a0a628} executeQuery: (from 172.16.1.1:64362) (query 1, line 1) CREATE TABLE table8 (`x` int) ENGINE = MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', `table` = 'mysql_table', user = 'mysql_user', password = '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.711364 [ 9 ] {44018ed3-7ebe-473a-87d9-50d699a0a628} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table8 (`x` int) ENGINE = MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', `table` = 'mysql_table', user = 'mysql_user', password = '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table9 (\\`x\\` int) ENGINE = MySQL(named_collection_2, database = \'mysql_db\', host = \'mysql80\', port = 3306, password = \'\\[HIDDEN\\]\', \\`table\\` = \'mysql_table\', user = \'mysql_user\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table9 (\`x\` int) ENGINE = MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = '\[HIDDEN\]', \`table\` = 'mysql_table', user = 'mysql_user')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.982419 [ 9 ] {e376c8c2-7f0f-42e6-9da7-a4cd09d30c39} executeQuery: (from 172.16.1.1:64368) (query 1, line 1) CREATE TABLE table9 (`x` int) ENGINE = MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = '[HIDDEN]', `table` = 'mysql_table', user = 'mysql_user') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:43.982616 [ 9 ] {e376c8c2-7f0f-42e6-9da7-a4cd09d30c39} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table9 (`x` int) ENGINE = MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = '[HIDDEN]', `table` = 'mysql_table', user = 'mysql_user') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = \'mysql_db\', host = \'mysql80\', port = 3306, table = \'mysql_table\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:44.240512 [ 9 ] {43b2a0b0-5a28-45c2-a4a7-dd7c05f9d9dc} executeQuery: (from 172.16.1.1:64380) (query 1, line 1) CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:44.240603 [ 9 ] {43b2a0b0-5a28-45c2-a4a7-dd7c05f9d9dc} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table11 (\\`x\\` int) ENGINE = PostgreSQL(named_collection_4, host = \'postgres1\', port = 5432, database = \'postgres_db\', \\`table\\` = \'postgres_table\', user = \'postgres_user\', password = \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table11 (\`x\` int) ENGINE = PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', \`table\` = 'postgres_table', user = 'postgres_user', password = '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:44.525782 [ 9 ] {38ca26c4-a819-45ec-a83b-f8eef0e0fe9b} executeQuery: (from 172.16.1.1:64396) (query 1, line 1) CREATE TABLE table11 (`x` int) ENGINE = PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', `table` = 'postgres_table', user = 'postgres_user', password = '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:44.525919 [ 9 ] {38ca26c4-a819-45ec-a83b-f8eef0e0fe9b} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table11 (`x` int) ENGINE = PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', `table` = 'postgres_table', user = 'postgres_user', password = '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table12 (\\`x\\` int) ENGINE = MongoDB(named_collection_5, host = \'mongo1\', port = 5432, database = \'mongo_db\', collection = \'mongo_col\', user = \'mongo_user\', password = \'\\[HIDDEN\\]\'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table12 (\`x\` int) ENGINE = MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = '\[HIDDEN\]'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:44.787082 [ 9 ] {6404cf72-5540-45c5-b328-d11e60f9c7f3} executeQuery: (from 172.16.1.1:64412) (query 1, line 1) CREATE TABLE table12 (`x` int) ENGINE = MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:44.787237 [ 9 ] {6404cf72-5540-45c5-b328-d11e60f9c7f3} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table12 (`x` int) ENGINE = MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table13 (\\`x\\` int) ENGINE = S3(named_collection_6, url = \'http://minio1:9001/root/data/test8.csv\', access_key_id = \'minio\', secret_access_key = \'\\[HIDDEN\\]\', format = \'CSV\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table13 (\`x\` int) ENGINE = S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = '\[HIDDEN\]', format = 'CSV')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.047717 [ 9 ] {84506989-d004-4bab-bfd4-159a31a594f6} executeQuery: (from 172.16.1.1:64426) (query 1, line 1) CREATE TABLE table13 (`x` int) ENGINE = S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = '[HIDDEN]', format = 'CSV') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.047810 [ 9 ] {84506989-d004-4bab-bfd4-159a31a594f6} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table13 (`x` int) ENGINE = S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = '[HIDDEN]', format = 'CSV') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table14 (x int) ENGINE = S3(\'http://minio1:9001/root/data/test9.csv.gz\', \'NOSIGN\', \'CSV\', \'gzip\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table14 (x int) ENGINE = S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.303294 [ 9 ] {146507a6-9273-4ff7-aa7d-22847e8fd477} executeQuery: (from 172.16.1.1:64430) (query 1, line 1) CREATE TABLE table14 (x int) ENGINE = S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.303473 [ 9 ] {146507a6-9273-4ff7-aa7d-22847e8fd477} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table14 (x int) ENGINE = S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table15 (\\`x\\` int) ENGINE = S3(\'http://minio1:9001/root/data/test10.csv.gz\', \'minio\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table15 (\`x\` int) ENGINE = S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.578167 [ 9 ] {afb0806b-111a-4dc0-8cd0-b833d328bfaa} executeQuery: (from 172.16.1.1:64438) (query 1, line 1) CREATE TABLE table15 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.578238 [ 9 ] {afb0806b-111a-4dc0-8cd0-b833d328bfaa} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table15 (`x` int) ENGINE = S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table16 (\\`x\\` int) ENGINE = DeltaLake(\'http://minio1:9001/root/data/test11.csv.gz\', \'minio\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table16 (\`x\` int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.847679 [ 9 ] {77e2fb2a-0732-4670-b118-823f64a9884c} executeQuery: (from 172.16.1.1:64448) (query 1, line 1) CREATE TABLE table16 (`x` int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:19:45.847860 [ 9 ] {77e2fb2a-0732-4670-b118-823f64a9884c} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table16 (`x` int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:00.386777 [ 9 ] {77e2fb2a-0732-4670-b118-823f64a9884c} executeQuery: Code: 499. DB::Exception: Failed to check existence of key data/test11.csv.gz/_delta_log/_last_checkpoint in bucket root: Code: 198. DB::NetException: Not found address of host: minio1. (DNS_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)). HTTP response code: 18446744073709551615, error type: NETWORK_CONNECTION. (S3_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (from 172.16.1.1:64448) (query 1, line 1) (in query: CREATE TABLE table16 (`x` int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')), Stack trace (when copying this message, always include the lines below): (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table17 (x int) ENGINE = S3Queue(\'http://minio1:9001/root/data/\', \'CSV\') settings mode = \'ordered\'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table17 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:00.684211 [ 9 ] {8ca25168-5eaa-4fe5-8820-5475370af49b} executeQuery: (from 172.16.1.1:64808) (query 1, line 1) CREATE TABLE table17 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered' (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:00.684326 [ 9 ] {8ca25168-5eaa-4fe5-8820-5475370af49b} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table17 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered' (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table18 (x int) ENGINE = S3Queue(\'http://minio1:9001/root/data/\', \'CSV\', \'gzip\') settings mode = \'ordered\'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table18 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.002521 [ 9 ] {449f4497-a56b-43e6-b513-b5b3af118cbd} executeQuery: (from 172.16.1.1:64812) (query 1, line 1) CREATE TABLE table18 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered' (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.002657 [ 9 ] {449f4497-a56b-43e6-b513-b5b3af118cbd} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table18 (x int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered' (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table19 (\\`x\\` int) ENGINE = S3Queue(\'http://minio1:9001/root/data/\', \'minio\', \'\\[HIDDEN\\]\', \'CSV\') SETTINGS mode = \'ordered\'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table19 (\`x\` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '\[HIDDEN\]', 'CSV') SETTINGS mode = 'ordered'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.305055 [ 9 ] {f8c28957-58da-4005-af04-5a2fe03ea706} executeQuery: (from 172.16.1.1:64816) (query 1, line 1) CREATE TABLE table19 (`x` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '[HIDDEN]', 'CSV') SETTINGS mode = 'ordered' (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.305145 [ 9 ] {f8c28957-58da-4005-af04-5a2fe03ea706} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table19 (`x` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '[HIDDEN]', 'CSV') SETTINGS mode = 'ordered' (cluster.py:145, run_and_check) 2025-06-13 14:21:20 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table20 (\\`x\\` int) ENGINE = S3Queue(\'http://minio1:9001/root/data/\', \'minio\', \'\\[HIDDEN\\]\', \'CSV\', \'gzip\') SETTINGS mode = \'ordered\'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:20 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table20 (\`x\` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '\[HIDDEN\]', 'CSV', 'gzip') SETTINGS mode = 'ordered'" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.626020 [ 9 ] {24d3cd39-a4d0-44ae-aa55-efd9b451add7} executeQuery: (from 172.16.1.1:64828) (query 1, line 1) CREATE TABLE table20 (`x` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '[HIDDEN]', 'CSV', 'gzip') SETTINGS mode = 'ordered' (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.626145 [ 9 ] {24d3cd39-a4d0-44ae-aa55-efd9b451add7} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table20 (`x` int) ENGINE = S3Queue('http://minio1:9001/root/data/', 'minio', '[HIDDEN]', 'CSV', 'gzip') SETTINGS mode = 'ordered' (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table21 (\\`x\\` int) ENGINE = Iceberg(\'http://minio1:9001/root/data/test11.csv.gz\', \'minio\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table21 (\`x\` int) ENGINE = Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.919227 [ 9 ] {d7c41f3b-85c1-4af0-a86c-5e0f6a05fed8} executeQuery: (from 172.16.1.1:64836) (query 1, line 1) CREATE TABLE table21 (`x` int) ENGINE = Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:01.919327 [ 9 ] {d7c41f3b-85c1-4af0-a86c-5e0f6a05fed8} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table21 (`x` int) ENGINE = Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:14.914423 [ 9 ] {d7c41f3b-85c1-4af0-a86c-5e0f6a05fed8} executeQuery: Code: 499. DB::Exception: Code: 198. DB::NetException: Not found address of host: minio1. (DNS_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (Code: 99, S3 exception: ''). (S3_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (from 172.16.1.1:64836) (query 1, line 1) (in query: CREATE TABLE table21 (`x` int) ENGINE = Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')), Stack trace (when copying this message, always include the lines below): (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table22 (\\`x\\` int) ENGINE = IcebergS3(\'http://minio1:9001/root/data/test11.csv.gz\', \'minio\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table22 (\`x\` int) ENGINE = IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:15.197544 [ 9 ] {84f6fc0d-ec38-4c14-aeb0-ba40bd6186c3} executeQuery: (from 172.16.1.1:60064) (query 1, line 1) CREATE TABLE table22 (`x` int) ENGINE = IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:15.197633 [ 9 ] {84f6fc0d-ec38-4c14-aeb0-ba40bd6186c3} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table22 (`x` int) ENGINE = IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:29.696225 [ 9 ] {84f6fc0d-ec38-4c14-aeb0-ba40bd6186c3} executeQuery: Code: 499. DB::Exception: Code: 198. DB::NetException: Not found address of host: minio1. (DNS_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (Code: 99, S3 exception: ''). (S3_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (from 172.16.1.1:60064) (query 1, line 1) (in query: CREATE TABLE table22 (`x` int) ENGINE = IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')), Stack trace (when copying this message, always include the lines below): (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table23 (\\`x\\` int) ENGINE = Iceberg(storage_type = \'s3\', \'http://minio1:9001/root/data/test11.csv.gz\', \'minio\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table23 (\`x\` int) ENGINE = Iceberg(storage_type = 's3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:29.961686 [ 9 ] {e82cb743-53f3-4409-bf26-9126c5b8961a} executeQuery: (from 172.16.1.1:63658) (query 1, line 1) CREATE TABLE table23 (`x` int) ENGINE = Iceberg(storage_type = 's3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:29.961900 [ 9 ] {e82cb743-53f3-4409-bf26-9126c5b8961a} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table23 (`x` int) ENGINE = Iceberg(storage_type = 's3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:44.771107 [ 9 ] {e82cb743-53f3-4409-bf26-9126c5b8961a} executeQuery: Code: 499. DB::Exception: Code: 198. DB::NetException: Not found address of host: minio1. (DNS_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (Code: 99, S3 exception: ''). (S3_ERROR) (version 25.3.3.20286.altinityantalya (altinity build)) (from 172.16.1.1:63658) (query 1, line 1) (in query: CREATE TABLE table23 (`x` int) ENGINE = Iceberg(storage_type = 's3', 'http://minio1:9001/root/data/test11.csv.gz', 'minio', '[HIDDEN]')), Stack trace (when copying this message, always include the lines below): (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table24 (\\`x\\` int) ENGINE = AzureBlobStorage(\'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\\[HIDDEN\\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;\', \'cont\', \'test_simple.csv\', \'CSV\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table24 (\`x\` int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\[HIDDEN\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple.csv', 'CSV')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.024556 [ 9 ] {63732e71-8fd3-4456-80dc-d6a8592fe91c} executeQuery: (from 172.16.1.1:57382) (query 1, line 1) CREATE TABLE table24 (`x` int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=[HIDDEN];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple.csv', 'CSV') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.024655 [ 9 ] {63732e71-8fd3-4456-80dc-d6a8592fe91c} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table24 (`x` int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=[HIDDEN];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple.csv', 'CSV') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table25 (\\`x\\` int) ENGINE = AzureBlobStorage(\'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\\[HIDDEN\\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;\', \'cont\', \'test_simple_1.csv\', \'CSV\', \'none\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table25 (\`x\` int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\[HIDDEN\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple_1.csv', 'CSV', 'none')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.309271 [ 9 ] {accd5664-8b19-4573-af9a-55235f17fc05} executeQuery: (from 172.16.1.1:57392) (query 1, line 1) CREATE TABLE table25 (`x` int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=[HIDDEN];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple_1.csv', 'CSV', 'none') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.309432 [ 9 ] {accd5664-8b19-4573-af9a-55235f17fc05} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table25 (`x` int) ENGINE = AzureBlobStorage('DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=[HIDDEN];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', 'cont', 'test_simple_1.csv', 'CSV', 'none') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table26 (\\`x\\` int) ENGINE = AzureBlobStorage(\'http://azurite1:30000/devstoreaccount1\', \'cont\', \'test_simple_3.csv\', \'devstoreaccount1\', \'\\[HIDDEN\\]\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table26 (\`x\` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_3.csv', 'devstoreaccount1', '\[HIDDEN\]')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.552067 [ 9 ] {92146fd9-c780-48bb-8dca-092ea7d7fd07} executeQuery: (from 172.16.1.1:57400) (query 1, line 1) CREATE TABLE table26 (`x` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_3.csv', 'devstoreaccount1', '[HIDDEN]') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.552170 [ 9 ] {92146fd9-c780-48bb-8dca-092ea7d7fd07} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table26 (`x` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_3.csv', 'devstoreaccount1', '[HIDDEN]') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table27 (\\`x\\` int) ENGINE = AzureBlobStorage(\'http://azurite1:30000/devstoreaccount1\', \'cont\', \'test_simple_4.csv\', \'devstoreaccount1\', \'\\[HIDDEN\\]\', \'CSV\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table27 (\`x\` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_4.csv', 'devstoreaccount1', '\[HIDDEN\]', 'CSV')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.881688 [ 9 ] {d633bc4a-7848-4f3b-a384-6223099eca63} executeQuery: (from 172.16.1.1:57414) (query 1, line 1) CREATE TABLE table27 (`x` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_4.csv', 'devstoreaccount1', '[HIDDEN]', 'CSV') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:45.881905 [ 9 ] {d633bc4a-7848-4f3b-a384-6223099eca63} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table27 (`x` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_4.csv', 'devstoreaccount1', '[HIDDEN]', 'CSV') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table28 (\\`x\\` int) ENGINE = AzureBlobStorage(\'http://azurite1:30000/devstoreaccount1\', \'cont\', \'test_simple_5.csv\', \'devstoreaccount1\', \'\\[HIDDEN\\]\', \'CSV\', \'none\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table28 (\`x\` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_5.csv', 'devstoreaccount1', '\[HIDDEN\]', 'CSV', 'none')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:46.155560 [ 9 ] {492f57de-a48b-4124-a0f3-2db528894333} executeQuery: (from 172.16.1.1:57416) (query 1, line 1) CREATE TABLE table28 (`x` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_5.csv', 'devstoreaccount1', '[HIDDEN]', 'CSV', 'none') (stage: Complete) (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.06.13 14:20:46.155634 [ 9 ] {492f57de-a48b-4124-a0f3-2db528894333} CancellationChecker: Did not add the task because the timeout is 0. Query: CREATE TABLE table28 (`x` int) ENGINE = AzureBlobStorage('http://azurite1:30000/devstoreaccount1', 'cont', 'test_simple_5.csv', 'devstoreaccount1', '[HIDDEN]', 'CSV', 'none') (cluster.py:145, run_and_check) 2025-06-13 14:21:21 [ 671 ] DEBUG : run container_id:roottestmasksensitiveinfo-gw0-node-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table29 (\\`x\\` int) ENGINE = AzureBlobStorage(named_collection_2, connection_string = \'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\\[HIDDEN\\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;\', container = \'cont\', blob_path = \'test_simple_7.csv\', format = \'CSV\')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true'] (cluster.py:2045, exec_in_container) 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker exec roottestmasksensitiveinfo-gw0-node-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CREATE TABLE table29 (\`x\` int) ENGINE = AzureBlobStorage(named_collection_2, connection_string = 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=\[HIDDEN\];BlobEndpoint=http://azurite1:30000/devstoreaccount1;', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "SELECT COUNT() FROM system.query_log WHERE query LIKE " ] && cat || grep -v "$SELECT COUNT() FROM system.query_log WHERE query LIKE " ) || true] (cluster.py:121, run_and_check) ---------------------------- Captured log teardown ----------------------------- 2025-06-13 14:21:21 [ 671 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml stop --timeout 20] (cluster.py:121, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml down --volumes] (cluster.py:121, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopped (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Volume roottestmasksensitiveinfo-gw0_data1-1 Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Removing (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Volume roottestmasksensitiveinfo-gw0_data1-1 Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Removed (cluster.py:147, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Cleanup called (cluster.py:846, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : Docker networks for project roottestmasksensitiveinfo-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:825, print_all_docker_pieces) 2025-06-13 14:21:39 [ 671 ] DEBUG : Docker containers for project roottestmasksensitiveinfo-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:833, print_all_docker_pieces) 2025-06-13 14:21:39 [ 671 ] DEBUG : Docker volumes for project roottestmasksensitiveinfo-gw0 are DRIVER VOLUME NAME (cluster.py:841, print_all_docker_pieces) 2025-06-13 14:21:39 [ 671 ] DEBUG : Command:[docker container list --all --filter name='^/roottestmasksensitiveinfo-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Unstopped containers: {} (cluster.py:860, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : No running containers for project: roottestmasksensitiveinfo-gw0 (cluster.py:874, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : Trying to prune unused networks... (cluster.py:880, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : Trying to prune unused images... (cluster.py:896, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Images pruned (cluster.py:899, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : Trying to prune unused volumes... (cluster.py:905, cleanup) 2025-06-13 14:21:39 [ 671 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2025-06-13 14:21:39 [ 671 ] DEBUG : Volumes pruned: 1 (cluster.py:910, cleanup) ----------------- generated report log file: parallel1_1.jsonl ----------------- ============================== slowest durations =============================== 100.36s call test_mask_sensitive_info/test.py::test_create_table 70.46s setup test_mask_sensitive_info/test.py::test_create_table 17.89s teardown test_mask_sensitive_info/test.py::test_create_table =========================== short test summary info ============================ FAILED test_mask_sensitive_info/test.py::test_create_table - assert False ======================== 1 failed in 191.36s (0:03:11) ========================= Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 492, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_typopj --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests test_mask_sensitive_info/test.py::test_create_table -vvv " altinityinfra/integration-tests-runner:ad96270260ff ' returned non-zero exit status 1.